## START: Set by rpmautospec ## (rpmautospec version 0.7.3) ## RPMAUTOSPEC: autorelease, autochangelog %define autorelease(e:s:pb:n) %{?-p:0.}%{lua: release_number = 4; base_release_number = tonumber(rpm.expand("%{?-b*}%{!?-b:1}")); print(release_number + base_release_number - 1); }%{?-e:.%{-e*}}%{?-s:.%{-s*}}%{!?-n:%{?dist}} ## END: Set by rpmautospec %global pypi_name torch # Where the src comes from %global forgeurl https://github.com/pytorch/pytorch # So pre releases can be tried %bcond_with gitcommit %if %{with gitcommit} # v2.5.0-rc9 %global commit0 417a0763a7d69f6ce80719ac89c1d2deeee78163 %global shortcommit0 %(c=%{commit0}; echo ${c:0:7}) %global date0 2024103 %global pypi_version 2.5.0 %else %global pypi_version 2.5.1 %endif # For -test subpackage # suitable only for local testing # Install and do something like # export LD_LIBRARY_PATH=/usr/lib64/python3.12/site-packages/torch/lib # /usr/lib64/python3.12/site-packages/torch/bin/test_api, test_lazy %bcond_with test %ifarch x86_64 %bcond_without rocm %endif %bcond_with rocm_loop %global rocm_default_gpu default %global rocm_gpu_list gfx9 # For testing distributed+rccl etc. %bcond_with rccl %bcond_with gloo %bcond_without mpi %bcond_without tensorpipe # Disable dwz with rocm because memory can be exhausted %if %{with rocm} %define _find_debuginfo_dwz_opts %{nil} %endif # These came in 2.4 and not yet in Fedora %bcond_with opentelemetry %bcond_with httplib %bcond_with kineto Name: python-%{pypi_name} %if %{with gitcommit} Version: %{pypi_version}^git%{date0}.%{shortcommit0} %else Version: %{pypi_version} %endif Release: %autorelease Summary: PyTorch AI/ML framework # See license.txt for license details License: BSD-3-Clause AND BSD-2-Clause AND 0BSD AND Apache-2.0 AND MIT AND BSL-1.0 AND GPL-3.0-or-later AND Zlib URL: https://pytorch.org/ %if %{with gitcommit} Source0: %{forgeurl}/archive/%{commit0}/pytorch-%{shortcommit0}.tar.gz Source1000: pyproject.toml %else Source0: %{forgeurl}/releases/download/v%{version}/pytorch-v%{version}.tar.gz %endif Source1: https://github.com/google/flatbuffers/archive/refs/tags/v23.3.3.tar.gz Source2: https://github.com/pybind/pybind11/archive/refs/tags/v2.11.1.tar.gz # Developement on tensorpipe has stopped, repo made read only July 1, 2023, this is the last commit %global tp_commit 52791a2fd214b2a9dc5759d36725909c1daa7f2e %global tp_scommit %(c=%{tp_commit}; echo ${c:0:7}) Source20: https://github.com/pytorch/tensorpipe/archive/%{tp_commit}/tensorpipe-%{tp_scommit}.tar.gz # The old libuv tensorpipe uses Source21: https://github.com/libuv/libuv/archive/refs/tags/v1.41.0.tar.gz # Developement afaik on libnop has stopped, this is the last commit %global nop_commit 910b55815be16109f04f4180e9adee14fb4ce281 %global nop_scommit %(c=%{nop_commit}; echo ${c:0:7}) Source22: https://github.com/google/libnop/archive/%{nop_commit}/libnop-%{nop_scommit}.tar.gz %if %{without opentelemetry} %global ot_ver 1.14.2 Source60: https://github.com/open-telemetry/opentelemetry-cpp/archive/refs/tags/v%{ot_ver}.tar.gz %endif %if %{without httplib} %global hl_commit 3b6597bba913d51161383657829b7e644e59c006 %global hl_scommit %(c=%{hl_commit}; echo ${c:0:7}) Source70: https://github.com/yhirose/cpp-httplib/archive/%{hl_commit}/cpp-httplib-%{hl_scommit}.tar.gz %endif %if %{without kineto} %global ki_commit be1317644c68b4bfc4646024a6b221066e430031 %global ki_scommit %(c=%{ki_commit}; echo ${c:0:7}) Source80: https://github.com/pytorch/kineto/archive/%{ki_commit}/kineto-%{ki_scommit}.tar.gz %endif Patch11: 0001-Improve-finding-and-using-the-rocm_version.h.patch # ROCm patches # Patches need to be refactored for ToT # These are ROCm packages Patch101: 0001-cuda-hip-signatures.patch ExclusiveArch: x86_64 aarch64 %global toolchain gcc %global _lto_cflags %nil BuildRequires: cmake BuildRequires: binutils-gold BuildRequires: eigen3-devel BuildRequires: flexiblas-devel BuildRequires: fmt-devel BuildRequires: foxi-devel BuildRequires: gcc-c++ BuildRequires: gcc-gfortran %if %{with gloo} BuildRequires: gloo-devel %endif BuildRequires: json-devel BuildRequires: libomp-devel BuildRequires: numactl-devel BuildRequires: ninja-build BuildRequires: onnx-devel %if %{with mpi} BuildRequires: openmpi-devel %endif BuildRequires: protobuf-devel BuildRequires: sleef-devel BuildRequires: valgrind-devel BuildRequires: pocketfft-devel BuildRequires: pthreadpool-devel BuildRequires: cpuinfo-devel BuildRequires: FP16-devel BuildRequires: fxdiv-devel BuildRequires: psimd-devel BuildRequires: xnnpack-devel = 0.0^git20240814.312eb7e BuildRequires: python3-devel BuildRequires: python3dist(filelock) BuildRequires: python3dist(jinja2) BuildRequires: python3dist(networkx) BuildRequires: python3dist(numpy) BuildRequires: python3dist(pyyaml) BuildRequires: python3dist(setuptools) BuildRequires: python3dist(sphinx) BuildRequires: python3dist(typing-extensions) %if 0%{?fedora} BuildRequires: python3-pybind11 BuildRequires: python3dist(fsspec) BuildRequires: python3dist(sympy) %endif %if %{with rocm} BuildRequires: hipblas-devel BuildRequires: hipblaslt-devel BuildRequires: hipcub-devel BuildRequires: hipfft-devel BuildRequires: hiprand-devel BuildRequires: hipsparse-devel BuildRequires: hipsolver-devel BuildRequires: magma-devel BuildRequires: miopen-devel BuildRequires: rocblas-devel BuildRequires: rocrand-devel BuildRequires: rocfft-devel %if %{with rccl} BuildRequires: rccl-devel %endif BuildRequires: rocprim-devel BuildRequires: rocm-cmake BuildRequires: rocm-comgr-devel BuildRequires: rocm-compilersupport-macros BuildRequires: rocm-core-devel BuildRequires: rocm-hip-devel BuildRequires: rocm-runtime-devel BuildRequires: rocm-rpm-macros BuildRequires: rocm-rpm-macros-modules BuildRequires: rocthrust-devel BuildRequires: roctracer-devel Requires: amdsmi Requires: rocm-rpm-macros-modules %endif %if %{with test} BuildRequires: google-benchmark-devel %endif Requires: python3dist(dill) Obsoletes: caffe = 1.0^git20200212.9b89154 %description PyTorch is a Python package that provides two high-level features: * Tensor computation (like NumPy) with strong GPU acceleration * Deep neural networks built on a tape-based autograd system You can reuse your favorite Python packages such as NumPy, SciPy, and Cython to extend PyTorch when needed. %package -n python3-%{pypi_name} Summary: %{summary} # For convience Provides: pytorch # Apache-2.0 Provides: bundled(flatbuffers) = 22.3.3 # MIT Provides: bundled(miniz) = 2.1.0 Provides: bundled(pybind11) = 2.11.1 %if %{with tensorpipe} # BSD-3-Clause Provides: bundled(tensorpipe) # Apache-2.0 Provides: bundled(libnop) # MIT AND CC-BY-4.0 AND ISC AND BSD-2-Clause Provides: bundled(libuv) = 1.41.0 %endif %description -n python3-%{pypi_name} PyTorch is a Python package that provides two high-level features: * Tensor computation (like NumPy) with strong GPU acceleration * Deep neural networks built on a tape-based autograd system You can reuse your favorite Python packages such as NumPy, SciPy, and Cython to extend PyTorch when needed. %if %{with rocm_loop} %package -n python3-%{pypi_name}-rocm-gfx9 Summary: %{name} for ROCm gfx9 %description -n python3-%{pypi_name}-rocm-gfx9 %{summary} %endif %if %{with test} %package -n python3-%{pypi_name}-test Summary: Tests for %{name} Requires: python3-%{pypi_name}%{?_isa} = %{version}-%{release} %description -n python3-%{pypi_name}-test %{summary} %endif %prep %if %{with gitcommit} %autosetup -p1 -n pytorch-%{commit0} # Overwrite with a git checkout of the pyproject.toml cp %{SOURCE1000} . %else %autosetup -p1 -n pytorch-v%{version} %endif # Remove bundled egg-info rm -rf %{pypi_name}.egg-info tar xf %{SOURCE1} rm -rf third_party/flatbuffers/* cp -r flatbuffers-23.3.3/* third_party/flatbuffers/ tar xf %{SOURCE2} rm -rf third_party/pybind11/* cp -r pybind11-2.11.1/* third_party/pybind11/ %if %{with tensorpipe} tar xf %{SOURCE20} rm -rf third_party/tensorpipe/* cp -r tensorpipe-*/* third_party/tensorpipe/ tar xf %{SOURCE21} rm -rf third_party/tensorpipe/third_party/libuv/* cp -r libuv-*/* third_party/tensorpipe/third_party/libuv/ tar xf %{SOURCE22} rm -rf third_party/tensorpipe/third_party/libnop/* cp -r libnop-*/* third_party/tensorpipe/third_party/libnop/ %endif %if %{without opentelemtry} tar xf %{SOURCE60} rm -rf third_party/opentelemetry-cpp/* cp -r opentelemetry-cpp-*/* third_party/opentelemetry-cpp/ %endif %if %{without httplib} tar xf %{SOURCE70} rm -rf third_party/cpp-httplib/* cp -r cpp-httplib-*/* third_party/cpp-httplib/ %endif %if %{without kineto} tar xf %{SOURCE80} rm -rf third_party/kineto/* cp -r kineto-*/* third_party/kineto/ %endif # hipblaslt only building with gfx90a sed -i -e 's@"gfx90a", "gfx940", "gfx941", "gfx942"@"gfx90a"@' aten/src/ATen/native/cuda/Blas.cpp %if 0%{?rhel} # In RHEL but too old sed -i -e '/typing-extensions/d' setup.py # Need to pip these sed -i -e '/sympy/d' setup.py sed -i -e '/fsspec/d' setup.py %else # for 2.5.0 sed -i -e 's@sympy==1.13.1@sympy>=1.13.1@' setup.py %endif # A new dependency # Connected to USE_FLASH_ATTENTION, since this is off, do not need it sed -i -e '/aotriton.cmake/d' cmake/Dependencies.cmake # Compress hip sed -i -e 's@HIP_CLANG_FLAGS -fno-gpu-rdc@HIP_CLANG_FLAGS -fno-gpu-rdc --offload-compress@' cmake/Dependencies.cmake # No third_party fmt, use system sed -i -e 's@fmt::fmt-header-only@fmt@' CMakeLists.txt sed -i -e 's@fmt::fmt-header-only@fmt@' c10/CMakeLists.txt sed -i -e 's@fmt::fmt-header-only@fmt@' torch/CMakeLists.txt sed -i -e 's@fmt::fmt-header-only@fmt@' cmake/Dependencies.cmake sed -i -e 's@fmt::fmt-header-only@fmt@' caffe2/CMakeLists.txt sed -i -e 's@add_subdirectory(${PROJECT_SOURCE_DIR}/third_party/fmt)@#add_subdirectory(${PROJECT_SOURCE_DIR}/third_party/fmt)@' cmake/Dependencies.cmake sed -i -e 's@set_target_properties(fmt-header-only PROPERTIES INTERFACE_COMPILE_FEATURES "")@#set_target_properties(fmt-header-only PROPERTIES INTERFACE_COMPILE_FEATURES "")@' cmake/Dependencies.cmake sed -i -e 's@list(APPEND Caffe2_DEPENDENCY_LIBS fmt::fmt-header-only)@#list(APPEND Caffe2_DEPENDENCY_LIBS fmt::fmt-header-only)@' cmake/Dependencies.cmake # No third_party FXdiv sed -i -e 's@if(NOT TARGET fxdiv)@if(MSVC AND USE_XNNPACK)@' caffe2/CMakeLists.txt sed -i -e 's@TARGET_LINK_LIBRARIES(torch_cpu PRIVATE fxdiv)@#TARGET_LINK_LIBRARIES(torch_cpu PRIVATE fxdiv)@' caffe2/CMakeLists.txt # Disable the use of check_submodule's in the setup.py, we are a tarball, not a git repo sed -i -e 's@check_submodules()$@#check_submodules()@' setup.py # Release comes fully loaded with third party src # Remove what we can # # For 2.1 this is all but miniz-2.1.0 # Instead of building as a library, caffe2 reaches into # the third_party dir to compile the file. # mimiz is licensed MIT # https://github.com/richgel999/miniz/blob/master/LICENSE mv third_party/miniz-2.1.0 . # # setup.py depends on this script mv third_party/build_bundled.py . # Need the just untarred flatbuffers/flatbuffers.h mv third_party/flatbuffers . mv third_party/pybind11 . %if %{with tensorpipe} mv third_party/tensorpipe . %endif %if %{without opentelemetry} mv third_party/opentelemetry-cpp . %endif %if %{without httplib} mv third_party/cpp-httplib . %endif %if %{without kineto} mv third_party/kineto . %endif %if %{with test} mv third_party/googletest . %endif # Remove everything rm -rf third_party/* # Put stuff back mv build_bundled.py third_party mv miniz-2.1.0 third_party mv flatbuffers third_party mv pybind11 third_party %if %{with tensorpipe} mv tensorpipe third_party %endif %if %{without opentelemetry} mv opentelemetry-cpp third_party %endif %if %{without httplib} mv cpp-httplib third_party %endif %if %{without kineto} mv kineto third_party %endif %if %{with test} mv googletest third_party %endif # # Fake out pocketfft, and system header will be used mkdir third_party/pocketfft # # Use the system valgrind headers mkdir third_party/valgrind-headers cp %{_includedir}/valgrind/* third_party/valgrind-headers # Fix installing to /usr/lib64 sed -i -e 's@DESTINATION ${PYTHON_LIB_REL_PATH}@DESTINATION ${CMAKE_INSTALL_PREFIX}/${PYTHON_LIB_REL_PATH}@' caffe2/CMakeLists.txt # reenable foxi linking sed -i -e 's@list(APPEND Caffe2_DEPENDENCY_LIBS foxi_loader)@#list(APPEND Caffe2_DEPENDENCY_LIBS foxi_loader)@' cmake/Dependencies.cmake %if %{with rocm} # hipify ./tools/amd_build/build_amd.py # Fedora installs to /usr/include, not /usr/include/rocm-core sed -i -e 's@rocm-core/rocm_version.h@rocm_version.h@' aten/src/ATen/hip/tunable/TunableGemm.h # use any hip, correct CMAKE_MODULE_PATH sed -i -e 's@lib/cmake/hip@lib64/cmake/hip@' cmake/public/LoadHIP.cmake sed -i -e 's@HIP 1.0@HIP MODULE@' cmake/public/LoadHIP.cmake # silence an assert # sed -i -e '/qvalue = std::clamp(qvalue, qmin, qmax);/d' aten/src/ATen/native/cuda/IndexKernel.cu %endif %build # # Control the number of jobs # # The build can fail if too many threads exceed the physical memory # So count core and and memory and increase the build memory util the build succeeds # # Real cores, No hyperthreading COMPILE_JOBS=`cat /proc/cpuinfo | grep -m 1 'cpu cores' | awk '{ print $4 }'` if [ ${COMPILE_JOBS}x = x ]; then COMPILE_JOBS=1 fi # Take into account memmory usage per core, do not thrash real memory BUILD_MEM=2 MEM_KB=0 MEM_KB=`cat /proc/meminfo | grep MemTotal | awk '{ print $2 }'` MEM_MB=`eval "expr ${MEM_KB} / 1024"` MEM_GB=`eval "expr ${MEM_MB} / 1024"` COMPILE_JOBS_MEM=`eval "expr 1 + ${MEM_GB} / ${BUILD_MEM}"` if [ "$COMPILE_JOBS_MEM" -lt "$COMPILE_JOBS" ]; then COMPILE_JOBS=$COMPILE_JOBS_MEM fi export MAX_JOBS=$COMPILE_JOBS # For debugging setup.py # export SETUPTOOLS_SCM_DEBUG=1 # For verbose cmake output # export VERBOSE=ON # For verbose linking # export CMAKE_SHARED_LINKER_FLAGS=-Wl,--verbose # Manually set this hardening flag export CMAKE_EXE_LINKER_FLAGS=-pie export BUILD_CUSTOM_PROTOBUF=OFF export BUILD_NVFUSER=OFF export BUILD_SHARED_LIBS=ON export BUILD_TEST=OFF export CMAKE_BUILD_TYPE=RelWithDebInfo export CMAKE_FIND_PACKAGE_PREFER_CONFIG=ON export CAFFE2_LINK_LOCAL_PROTOBUF=OFF export INTERN_BUILD_MOBILE=OFF export USE_DISTRIBUTED=OFF export USE_CUDA=OFF export USE_FAKELOWP=OFF export USE_FBGEMM=OFF export USE_FLASH_ATTENTION=OFF export USE_GOLD_LINKER=ON export USE_GLOO=OFF export USE_ITT=OFF export USE_KINETO=OFF export USE_LITE_INTERPRETER_PROFILER=OFF export USE_LITE_PROTO=OFF export USE_MAGMA=OFF export USE_MEM_EFF_ATTENTION=OFF export USE_MKLDNN=OFF export USE_MPI=OFF export USE_NCCL=OFF export USE_NNPACK=OFF export USE_NUMPY=ON export USE_OPENMP=ON export USE_PYTORCH_QNNPACK=OFF export USE_ROCM=OFF export USE_SYSTEM_SLEEF=ON export USE_SYSTEM_EIGEN_INSTALL=ON export USE_SYSTEM_ONNX=ON export USE_SYSTEM_PYBIND11=OFF export USE_SYSTEM_LIBS=OFF export USE_TENSORPIPE=OFF export USE_XNNPACK=ON export USE_XPU=OFF export USE_SYSTEM_PTHREADPOOL=ON export USE_SYSTEM_CPUINFO=ON export USE_SYSTEM_FP16=ON export USE_SYSTEM_FXDIV=ON export USE_SYSTEM_PSIMD=ON export USE_SYSTEM_XNNPACK=ON export USE_DISTRIBUTED=ON %if %{with tensorpipe} export USE_TENSORPIPE=ON export TP_BUILD_LIBUV=OFF %endif %if %{with gloo} export USE_GLOO=ON export USE_SYSTEM_GLOO=ON %endif %if %{with mpi} export USE_MPI=ON %endif %if %{with test} export BUILD_TEST=ON %endif # Why we are using py3_ vs pyproject_ # # current pyproject problem with mock # + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_wheel.py /builddir/build/BUILD/pytorch-v2.1.0/pyproject-wheeldir # /usr/bin/python3: No module named pip # Adding pip to build requires does not fix # # See BZ 2244862 %if %{with rocm} export USE_ROCM=ON export USE_MAGMA=ON export HIP_PATH=`hipconfig -p` export ROCM_PATH=`hipconfig -R` RESOURCE_DIR=`%{rocmllvm_bindir}/clang -print-resource-dir` export DEVICE_LIB_PATH=${RESOURCE_DIR}/amdgcn/bitcode # pytorch uses clang, not hipcc export HIP_CLANG_PATH=%{rocmllvm_bindir} gpu=%{rocm_default_gpu} module load rocm/$gpu export PYTORCH_ROCM_ARCH=$ROCM_GPUS %py3_build mv build build-${gpu} module purge %if %{with rocm_loop} for gpu in %{rocm_gpu_list} do module load rocm/$gpu export PYTORCH_ROCM_ARCH=$ROCM_GPUS %py3_build mv build build-${gpu} module purge done %endif %else %py3_build %endif %install %if %{with rocm} export USE_ROCM=ON export HIP_PATH=`hipconfig -p` export ROCM_PATH=`hipconfig -R` RESOURCE_DIR=`%{rocmllvm_bindir}/clang -print-resource-dir` export DEVICE_LIB_PATH=${RESOURCE_DIR}/amdgcn/bitcode # pytorch uses clang, not hipcc export HIP_CLANG_PATH=%{rocmllvm_bindir} gpu=%{rocm_default_gpu} module load rocm/$gpu export PYTORCH_ROCM_ARCH=$ROCM_GPUS mv build-${gpu} build %py3_install mv build build-${gpu} module purge %if %{with rocm_loop} for gpu in %{rocm_gpu_list} do module load rocm/$gpu export PYTORCH_ROCM_ARCH=$ROCM_GPUS mv build-${gpu} build # need to customize the install location, so replace py3_install %{__python3} %{py_setup} %{?py_setup_args} install -O1 --skip-build --root %{buildroot} --prefix /usr/lib64/rocm/${gpu} %{?*} rm -rfv %{buildroot}/usr/lib/rocm/${gpu}/bin/__pycache__ mv build build-${gpu} module purge done %endif %else %py3_install %endif %check %py3_check_import torch # Do not remote the empty files %files -n python3-%{pypi_name} %license LICENSE %doc README.md %{_bindir}/convert-caffe2-to-onnx %{_bindir}/convert-onnx-to-caffe2 %{_bindir}/torchrun %{_bindir}/torchfrtrace %{python3_sitearch}/%{pypi_name} %{python3_sitearch}/%{pypi_name}-*.egg-info %{python3_sitearch}/functorch %{python3_sitearch}/torchgen %if %{with rocm} %if %{with rocm_loop} %files -n python3-%{pypi_name}-rocm-gfx9 %{_libdir}/rocm/gfx9/bin/* %{_libdir}/rocm/gfx9/lib64/* %endif %endif %changelog ## START: Generated by rpmautospec * Thu Jan 02 2025 Tom Rix - 2.5.1-4 - Rebuild for onnx * Tue Dec 24 2024 Tom Rix - 2.5.1-3 - Remove many options * Mon Dec 23 2024 Tom Rix - 2.5.1-2 - Obsolete caffe * Sat Dec 21 2024 Tom Rix - 2.5.1-1 - Update to 2.5.1 * Tue Nov 26 2024 Tom Rix - 2.5.0-1 - Update for 2.5.0 * Thu Nov 14 2024 Tom Rix - 2.4.1-12 - Use rocmllvm_bindir * Thu Oct 31 2024 Peter Robinson - 2.4.1-11 - drop old versions of pytorch from sources * Thu Oct 31 2024 Peter Robinson - 2.4.1-10 - Add binutils-gold build dep * Thu Oct 31 2024 Peter Robinson - 2.4.1-9 - Fix various Provides including the pytotch provides * Tue Oct 29 2024 Tom Rix - 2.4.1-8 - Use the new xnnpack * Fri Oct 11 2024 Tom Rix - 2.4.1-7 - Update gitcommit to v2.5.0-rc9 * Thu Oct 10 2024 Tom Rix - 2.4.1-6 - Update for llvm18 * Mon Oct 07 2024 Tom Rix - 2.4.1-5 - Some help finding llvm18 * Mon Sep 30 2024 Tom Rix - 2.4.1-4 - Update gitcommit * Sun Sep 15 2024 Tom Rix - 2.4.1-3 - Simplify cuda versions * Sun Sep 15 2024 Tom Rix - 2.4.1-2 - Update gitcommit * Mon Sep 09 2024 Tom Rix - 2.4.1-1 - Update to 2.4.1 * Tue Sep 03 2024 Tom Rix - 2.4.0-10 - amdsmi is a runtime dependency for ROCm * Fri Aug 30 2024 Tom Rix - 2.4.0-9 - Update the gitcommit * Thu Aug 15 2024 Tom Rix - 2.4.0-8 - Start tracking 2.5 * Wed Aug 07 2024 Tom Rix - 2.4.0-7 - Disable fbgemm with rocm * Mon Aug 05 2024 Tom Rix - 2.4.0-6 - Enable hipblaslt * Sun Aug 04 2024 Tom Rix - 2.4.0-5 - Remove the packages * Sun Aug 04 2024 Tom Rix - 2.4.0-4 - Simplify ROCm gpu list * Sat Jul 27 2024 Tom Rix - 2.4.0-3 - Fbgemm not available on aarch64 * Thu Jul 25 2024 Sérgio M. Basto - 2.4.0-2 - Rebuild for opencv 4.10.0 * Thu Jul 25 2024 Tom Rix - 2.4.0-1 - PyTorch 2.4 * Sat Jul 20 2024 Tom Rix - 2.3.1-23 - Fix USE_NUMA * Sat Jul 20 2024 Tom Rix - 2.3.1-22 - Use fbgemm on 2.4 * Tue Jul 16 2024 Kefu Chai - 2.3.1-20 - Rebuilt for fmt 11 * Wed Jul 10 2024 Tom Rix - 2.3.1-19 - Update to 2.4-rc8 * Fri Jul 05 2024 Tom Rix - 2.3.1-18 - Switch from openblas to flexiblas (rhbz#2295953) * Thu Jul 04 2024 Tom Rix - 2.3.1-17 - Show use of hipblaslt package * Thu Jul 04 2024 Tom Rix - 2.3.1-16 - Revisions of patches for 2.4 * Wed Jun 26 2024 Tom Rix - 2.3.1-15 - Add a CUDA subpackage * Wed Jun 26 2024 Tom Rix - 2.3.1-14 - Update gitcommit to v2.4.0-rc6 * Tue Jun 25 2024 Tom Rix - 2.3.1-13 - Add CUDA BuildRequires * Mon Jun 24 2024 Tom Rix - 2.3.1-12 - Update gitcommit to 2.4.0-rc5 * Fri Jun 21 2024 Tom Rix - 2.3.1-11 - Update gitcommit to 2.4.0-rc3 * Tue Jun 18 2024 Benjamin A. Beasley - 2.3.1-10 - Patch for sleef 3.6 * Fri Jun 14 2024 Python Maint - 2.3.1-9 - Rebuilt for Python 3.13 * Thu Jun 13 2024 Tom Rix - 2.3.1-8 - Update gitcommit * Thu Jun 13 2024 Tom Rix - 2.3.1-7 - Use specific version of CUDA base on disto release * Tue Jun 11 2024 Tom Rix - 2.3.1-6 - Fix broken cpuinfo for aarch64 * Tue Jun 11 2024 Tom Rix - 2.3.1-5 - Reduce amd gpu list on F40 * Mon Jun 10 2024 Tom Rix - 2.3.1-4 - Start a readme for NVIDIA * Mon Jun 10 2024 Tom Rix - 2.3.1-3 - Fix the normal build. * Sun Jun 09 2024 Tom Rix - 2.3.1-2 - Update gitcommit * Sun Jun 09 2024 Tom Rix - 2.3.1-1 - Update to 2.3.1 * Sat Jun 08 2024 Tom Rix - 2.3.0-15 - Add --with compat_gcc * Sat Jun 08 2024 Tom Rix - 2.3.0-14 - Do not apply ROCm patches with CUDA build * Fri Jun 07 2024 Tom Rix - 2.3.0-13 - Do not conditionally patch * Thu Jun 06 2024 Tom Rix - 2.3.0-12 - Update for ROCm 6.1.1 * Wed Jun 05 2024 Tom Rix - 2.3.0-11 - Update the ToT git commit * Tue May 21 2024 Tom Rix - 2.3.0-10 - Start tracking upstream 2.4 * Sat May 18 2024 Tom Rix - 2.3.0-9 - Roll ROCm support claim back to f40 * Thu May 16 2024 Tom Rix - 2.3.0-8 - Add cuda arches to build for * Tue May 07 2024 Tom Rix - 2.3.0-7 - Fill in missing packages on F40 and F39 with third_party. * Sun May 05 2024 Tom Rix - 2.3.0-6 - Collect the buildrequires that depend on F40 together. * Sun May 05 2024 Tom Rix - 2.3.0-5 - Improve fedora conditional use versions. * Fri May 03 2024 Tom Rix - 2.3.0-4 - Enable dynamo on 3.12 * Thu May 02 2024 Tom Rix - 2.3.0-3 - Disable dwz with ROCm * Tue Apr 30 2024 Tom Rix - 2.3.0-2 - Update sources * Tue Apr 30 2024 Tom Rix - 2.3.0-1 - Initial 2.3 release * Mon Apr 15 2024 Tom Rix - 2.3.0^git20240408.97ff6cf-2 - Use the system gloo * Thu Apr 11 2024 Tom Rix - 2.3.0^git20240408.97ff6cf-1 - v2.3.0-rc12 * Sat Apr 06 2024 Tom Rix - 2.3.0^git20240402.4bb5cb5-1 - Update to 2.3-rc7 * Sun Mar 31 2024 Tom Rix - 2.3.0^git20242213.74832f1-2 - Provide pytorch as a convience * Wed Mar 27 2024 Tom Rix - 2.3.0^git20242213.74832f1-1 - Update to 2.3-rc6 * Fri Mar 22 2024 Tom Rix - 2.3.0^git20240313.6a89a75-8 - Remove conditional around the rocm patches * Fri Mar 22 2024 Tom Rix - 2.3.0^git20240313.6a89a75-7 - Split the ROCm gpu families out into subpackages. * Thu Mar 21 2024 Tom Rix - 2.3.0^git20240313.6a89a75-6 - Update the source to 2.3-rc2 * Thu Mar 21 2024 Tom Rix - 2.3.0^git20240313.6a89a75-5 - RPMAUTOSPEC: unresolvable merge ## END: Generated by rpmautospec